2025-07-28 09:47:57,613 [ 323346 ] INFO : ClickHouse root is not set. Will use /home/ubuntu/_work/ClickHouse/ClickHouse (runner:53, check_args_and_update_paths) 2025-07-28 09:47:57,613 [ 323346 ] INFO : Cases dir is not set. Will use /home/ubuntu/_work/ClickHouse/ClickHouse/tests/integration (runner:79, check_args_and_update_paths) 2025-07-28 09:47:57,613 [ 323346 ] INFO : utils dir is not set. Will use /home/ubuntu/_work/ClickHouse/ClickHouse/utils (runner:90, check_args_and_update_paths) 2025-07-28 09:47:57,613 [ 323346 ] INFO : base_configs_dir: /home/ubuntu/_work/ClickHouse/ClickHouse/programs/server, binary: /home/ubuntu/_work/_temp/test/build/clickhouse, cases_dir: /home/ubuntu/_work/ClickHouse/ClickHouse/tests/integration (runner:92, check_args_and_update_paths) clickhouse_integration_tests_volume Running pytest container as: 'docker run --rm --name clickhouse_integration_tests_53fn1r --privileged --dns-search='.' --memory=30709018624 --security-opt seccomp=unconfined --cap-add=SYS_PTRACE --volume=/home/ubuntu/_work/_temp/test/build/clickhouse:/clickhouse --volume=/home/ubuntu/_work/ClickHouse/ClickHouse/programs/server:/clickhouse-config --volume=/home/ubuntu/_work/ClickHouse/ClickHouse/tests/integration:/ClickHouse/tests/integration --volume=/home/ubuntu/_work/ClickHouse/ClickHouse/utils/backupview:/ClickHouse/utils/backupview --volume=/home/ubuntu/_work/ClickHouse/ClickHouse/utils/grpc-client/pb2:/ClickHouse/utils/grpc-client/pb2 --volume=/run:/run/host:ro --volume=clickhouse_integration_tests_volume:/var/lib/docker -e DOCKER_DOTNET_CLIENT_TAG=11de0b29a15d -e DOCKER_HELPER_TAG=5dc43a6382f0 -e DOCKER_BASE_TAG=5ccda723c1fc -e DOCKER_KERBEROS_KDC_TAG=9391ecdee8d7 -e DOCKER_MYSQL_GOLANG_CLIENT_TAG=9bec2a638e6e -e DOCKER_MYSQL_JAVA_CLIENT_TAG=766bff31cfe4 -e DOCKER_MYSQL_JS_CLIENT_TAG=41ba7c2ec2a1 -e DOCKER_MYSQL_PHP_CLIENT_TAG=88be89c1e3b6 -e DOCKER_NGINX_DAV_TAG=b55ac9cd7519 -e DOCKER_POSTGRESQL_JAVA_CLIENT_TAG=a4eff5c7f4d6 -e DOCKER_PYTHON_BOTTLE_TAG=d862517635bf -e DOCKER_CLIENT_TIMEOUT=300 -e COMPOSE_HTTP_TIMEOUT=600 -e PYTHONUNBUFFERED=1 -e PYTEST_ADDOPTS="--dist=loadfile -n 10 -rfEps --run-id=1 --color=no --durations=0 --report-log=parallel1_1.jsonl --report-log-exclude-logs-on-passed-tests test_fetch_memory_usage/test.py::test_huge_column test_hedged_requests_parallel/test.py::test_combination1 test_hedged_requests_parallel/test.py::test_combination2 test_hedged_requests_parallel/test.py::test_query_with_no_data_to_sample test_hedged_requests_parallel/test.py::test_send_data test_hedged_requests_parallel/test.py::test_send_table_status_sleep test_jemalloc_percpu_arena/test.py::test_jemalloc_percpu_arena test_merge_tree_check_part_with_cache/test.py::test_check_part_with_cache test_merge_tree_load_parts/test.py::test_merge_tree_load_parts_filesystem_error test_overcommit_tracker/test.py::test_user_overcommit -vvv " altinityinfra/integration-tests-runner:226bfaf75ac1 '. Start tests ============================= test session starts ============================== platform linux -- Python 3.10.12, pytest-7.4.4, pluggy-1.5.0 -- /usr/bin/python3 cachedir: .pytest_cache Test order randomisation NOT enabled. Enable with --random-order or --random-order-bucket= rootdir: /ClickHouse/tests/integration configfile: pytest.ini plugins: timeout-2.3.1, repeat-0.9.3, order-1.0.0, reportlog-0.4.0, xdist-3.5.0, random-order-1.1.1 timeout: 900.0s timeout method: signal timeout func_only: False created: 10/10 workers 10 workers [10 items] scheduling tests via LoadFileScheduling test_fetch_memory_usage/test.py::test_huge_column test_overcommit_tracker/test.py::test_user_overcommit test_hedged_requests_parallel/test.py::test_combination1 test_merge_tree_load_parts/test.py::test_merge_tree_load_parts_filesystem_error test_merge_tree_check_part_with_cache/test.py::test_check_part_with_cache test_jemalloc_percpu_arena/test.py::test_jemalloc_percpu_arena [gw3] [ 10%] SKIPPED test_jemalloc_percpu_arena/test.py::test_jemalloc_percpu_arena [gw2] [ 20%] SKIPPED test_merge_tree_check_part_with_cache/test.py::test_check_part_with_cache [gw0] [ 30%] SKIPPED test_hedged_requests_parallel/test.py::test_combination1 test_hedged_requests_parallel/test.py::test_combination2 [gw0] [ 40%] SKIPPED test_hedged_requests_parallel/test.py::test_combination2 test_hedged_requests_parallel/test.py::test_query_with_no_data_to_sample [gw0] [ 50%] SKIPPED test_hedged_requests_parallel/test.py::test_query_with_no_data_to_sample test_hedged_requests_parallel/test.py::test_send_data [gw0] [ 60%] SKIPPED test_hedged_requests_parallel/test.py::test_send_data test_hedged_requests_parallel/test.py::test_send_table_status_sleep [gw0] [ 70%] SKIPPED test_hedged_requests_parallel/test.py::test_send_table_status_sleep [gw5] [ 80%] SKIPPED test_merge_tree_load_parts/test.py::test_merge_tree_load_parts_filesystem_error [gw1] [ 90%] SKIPPED test_fetch_memory_usage/test.py::test_huge_column [gw4] [100%] FAILED test_overcommit_tracker/test.py::test_user_overcommit =================================== FAILURES =================================== _____________________________ test_user_overcommit _____________________________ [gw4] linux -- Python 3.10.12 /usr/bin/python3 def test_user_overcommit(): node.query("CREATE USER IF NOT EXISTS A") node.query("GRANT ALL ON *.* TO A") responses_A = list() responses_B = list() for i in range(100): if i % 2 == 0: responses_A.append(node.get_query_request(USER_TEST_QUERY_A, user="A")) else: responses_B.append(node.get_query_request(USER_TEST_QUERY_B, user="A")) overcommited_killed = False for response in responses_A: _, err = response.get_answer_and_error() if "MEMORY_LIMIT_EXCEEDED" in err: overcommited_killed = True finished = False for response in responses_B: _, err = response.get_answer_and_error() if err == "": finished = True > assert finished, "all tasks are killed" E AssertionError: all tasks are killed E assert False test_overcommit_tracker/test.py:52: AssertionError ---------------------------- Captured stdout setup ----------------------------- Copy common default production configuration from /clickhouse-config. Files: config.xml, users.xml ------------------------------ Captured log setup ------------------------------ 2025-07-28 09:48:04.193000 [ 682 ] DEBUG : Command:[docker ps | wc -l] (cluster.py:121, run_and_check) 2025-07-28 09:48:04.210000 [ 682 ] DEBUG : Stdout:1 (cluster.py:145, run_and_check) 2025-07-28 09:48:04.210000 [ 682 ] DEBUG : No running containers (conftest.py:95, cleanup_environment) 2025-07-28 09:48:04.210000 [ 682 ] DEBUG : Pruning Docker networks (conftest.py:97, cleanup_environment) 2025-07-28 09:48:04.210000 [ 682 ] DEBUG : Command:[docker network prune --force] (cluster.py:121, run_and_check) 2025-07-28 09:48:04.230000 [ 682 ] DEBUG : Command:[sysctl net.ipv4.ip_local_port_range='55000 65535'] (cluster.py:121, run_and_check) 2025-07-28 09:48:04.233000 [ 682 ] DEBUG : Stdout:net.ipv4.ip_local_port_range = 55000 65535 (cluster.py:145, run_and_check) 2025-07-28 09:48:04.234000 [ 682 ] INFO : Running tests in /ClickHouse/tests/integration/test_overcommit_tracker/test.py (cluster.py:2738, start) 2025-07-28 09:48:04.234000 [ 682 ] DEBUG : Cluster start called. is_up=False (cluster.py:2745, start) 2025-07-28 09:48:04.254000 [ 682 ] DEBUG : Docker networks for project roottestovercommittracker-gw4 are NETWORK ID NAME DRIVER SCOPE (cluster.py:830, print_all_docker_pieces) 2025-07-28 09:48:04.272000 [ 682 ] DEBUG : Docker containers for project roottestovercommittracker-gw4 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES (cluster.py:838, print_all_docker_pieces) 2025-07-28 09:48:04.293000 [ 682 ] DEBUG : Docker volumes for project roottestovercommittracker-gw4 are DRIVER VOLUME NAME (cluster.py:846, print_all_docker_pieces) 2025-07-28 09:48:04.293000 [ 682 ] DEBUG : Cleanup called (cluster.py:851, cleanup) 2025-07-28 09:48:04.314000 [ 682 ] DEBUG : Docker networks for project roottestovercommittracker-gw4 are NETWORK ID NAME DRIVER SCOPE (cluster.py:830, print_all_docker_pieces) 2025-07-28 09:48:04.331000 [ 682 ] DEBUG : Docker containers for project roottestovercommittracker-gw4 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES (cluster.py:838, print_all_docker_pieces) 2025-07-28 09:48:04.348000 [ 682 ] DEBUG : Docker volumes for project roottestovercommittracker-gw4 are DRIVER VOLUME NAME (cluster.py:846, print_all_docker_pieces) 2025-07-28 09:48:04.348000 [ 682 ] DEBUG : Command:[docker container list --all --filter name='^/roottestovercommittracker-gw4-.*-1$' --format '{{.ID}}:{{.Names}}'] (cluster.py:121, run_and_check) 2025-07-28 09:48:04.366000 [ 682 ] DEBUG : Unstopped containers: {} (cluster.py:865, cleanup) 2025-07-28 09:48:04.366000 [ 682 ] DEBUG : No running containers for project: roottestovercommittracker-gw4 (cluster.py:879, cleanup) 2025-07-28 09:48:04.366000 [ 682 ] DEBUG : Trying to prune unused networks... (cluster.py:885, cleanup) 2025-07-28 09:48:04.385000 [ 682 ] DEBUG : Trying to prune unused images... (cluster.py:901, cleanup) 2025-07-28 09:48:04.385000 [ 682 ] DEBUG : Command:[docker image prune -f] (cluster.py:121, run_and_check) 2025-07-28 09:48:04.402000 [ 682 ] DEBUG : Stderr:Error response from daemon: a prune operation is already running (cluster.py:147, run_and_check) 2025-07-28 09:48:04.402000 [ 682 ] DEBUG : Exitcode:1 (cluster.py:149, run_and_check) 2025-07-28 09:48:04.402000 [ 682 ] DEBUG : Trying to prune unused volumes... (cluster.py:910, cleanup) 2025-07-28 09:48:04.403000 [ 682 ] DEBUG : Command:[docker volume ls | wc -l] (cluster.py:121, run_and_check) 2025-07-28 09:48:04.417000 [ 682 ] DEBUG : Stdout:1 (cluster.py:145, run_and_check) 2025-07-28 09:48:04.418000 [ 682 ] DEBUG : Volumes pruned: 1 (cluster.py:915, cleanup) 2025-07-28 09:48:04.418000 [ 682 ] DEBUG : Setup directory for instance: node (cluster.py:2758, start) 2025-07-28 09:48:04.419000 [ 682 ] DEBUG : Create directory for configuration generated in this helper (cluster.py:4628, create_dir) 2025-07-28 09:48:04.419000 [ 682 ] DEBUG : Create directory for common tests configuration (cluster.py:4633, create_dir) 2025-07-28 09:48:04.419000 [ 682 ] DEBUG : Copy common configuration from helpers (cluster.py:4653, create_dir) 2025-07-28 09:48:04.419000 [ 682 ] DEBUG : Generate and write macros file (cluster.py:4705, create_dir) 2025-07-28 09:48:04.420000 [ 682 ] DEBUG : Copy custom test config files ['/ClickHouse/tests/integration/test_overcommit_tracker/configs/config.d/config.xml'] to /ClickHouse/tests/integration/test_overcommit_tracker/_instances-1-gw4/node/configs/config.d (cluster.py:4741, create_dir) 2025-07-28 09:48:04.421000 [ 682 ] DEBUG : Setup database dir /ClickHouse/tests/integration/test_overcommit_tracker/_instances-1-gw4/node/database (cluster.py:4758, create_dir) 2025-07-28 09:48:04.421000 [ 682 ] DEBUG : Setup logs dir /ClickHouse/tests/integration/test_overcommit_tracker/_instances-1-gw4/node/logs (cluster.py:4769, create_dir) 2025-07-28 09:48:04.421000 [ 682 ] DEBUG : Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] (cluster.py:4850, create_dir) 2025-07-28 09:48:04.421000 [ 682 ] DEBUG : Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw'} stored in /ClickHouse/tests/integration/test_overcommit_tracker/_instances-1-gw4/.env (cluster.py:96, _create_env_file) 2025-07-28 09:48:04.422000 [ 682 ] DEBUG : Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] (config.py:21, find_config_file) 2025-07-28 09:48:04.422000 [ 682 ] DEBUG : No config file found (config.py:28, find_config_file) 2025-07-28 09:48:04.422000 [ 682 ] DEBUG : Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] (config.py:21, find_config_file) 2025-07-28 09:48:04.422000 [ 682 ] DEBUG : No config file found (config.py:28, find_config_file) 2025-07-28 09:48:04.430000 [ 682 ] DEBUG : http://localhost:None "GET /version HTTP/1.1" 200 826 (connectionpool.py:547, _make_request) 2025-07-28 09:48:04.430000 [ 682 ] DEBUG : Command:[docker compose --env-file /ClickHouse/tests/integration/test_overcommit_tracker/_instances-1-gw4/.env --project-name roottestovercommittracker-gw4 --file /ClickHouse/tests/integration/test_overcommit_tracker/_instances-1-gw4/node/docker-compose.yml pull] (cluster.py:121, run_and_check) 2025-07-28 09:48:15.088000 [ 682 ] DEBUG : Stderr: node Pulling (cluster.py:147, run_and_check) 2025-07-28 09:48:15.089000 [ 682 ] DEBUG : Stderr: node Pulled (cluster.py:147, run_and_check) 2025-07-28 09:48:15.089000 [ 682 ] DEBUG : ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_overcommit_tracker/_instances-1-gw4/.env --project-name roottestovercommittracker-gw4 --file /ClickHouse/tests/integration/test_overcommit_tracker/_instances-1-gw4/node/docker-compose.yml up -d --no-recreate') (cluster.py:3139, start) 2025-07-28 09:48:15.089000 [ 682 ] DEBUG : Command:[docker compose --env-file /ClickHouse/tests/integration/test_overcommit_tracker/_instances-1-gw4/.env --project-name roottestovercommittracker-gw4 --file /ClickHouse/tests/integration/test_overcommit_tracker/_instances-1-gw4/node/docker-compose.yml up -d --no-recreate] (cluster.py:121, run_and_check) 2025-07-28 09:48:16.321000 [ 682 ] DEBUG : Stderr: Network roottestovercommittracker-gw4_default Creating (cluster.py:147, run_and_check) 2025-07-28 09:48:16.321000 [ 682 ] DEBUG : Stderr: Network roottestovercommittracker-gw4_default Created (cluster.py:147, run_and_check) 2025-07-28 09:48:16.321000 [ 682 ] DEBUG : Stderr: Container roottestovercommittracker-gw4-node-1 Creating (cluster.py:147, run_and_check) 2025-07-28 09:48:16.322000 [ 682 ] DEBUG : Stderr: Container roottestovercommittracker-gw4-node-1 Created (cluster.py:147, run_and_check) 2025-07-28 09:48:16.322000 [ 682 ] DEBUG : Stderr: Container roottestovercommittracker-gw4-node-1 Starting (cluster.py:147, run_and_check) 2025-07-28 09:48:16.322000 [ 682 ] DEBUG : Stderr: Container roottestovercommittracker-gw4-node-1 Started (cluster.py:147, run_and_check) 2025-07-28 09:48:16.322000 [ 682 ] DEBUG : ClickHouse instance created (cluster.py:3147, start) 2025-07-28 09:48:16.322000 [ 682 ] DEBUG : get_instance_ip instance_name=node (cluster.py:2005, get_instance_ip) 2025-07-28 09:48:16.324000 [ 682 ] DEBUG : http://localhost:None "GET /v1.46/containers/roottestovercommittracker-gw4-node-1/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-07-28 09:48:16.325000 [ 682 ] DEBUG : get_instance_ip instance_name=node (cluster.py:2015, get_instance_global_ipv6) 2025-07-28 09:48:16.326000 [ 682 ] DEBUG : http://localhost:None "GET /v1.46/containers/roottestovercommittracker-gw4-node-1/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-07-28 09:48:16.327000 [ 682 ] DEBUG : Waiting for ClickHouse start in node, ip: 172.16.3.2... (cluster.py:3155, start) 2025-07-28 09:48:16.328000 [ 682 ] DEBUG : http://localhost:None "GET /v1.46/containers/roottestovercommittracker-gw4-node-1/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-07-28 09:48:16.330000 [ 682 ] DEBUG : http://localhost:None "GET /v1.46/containers/037cfbebda22295370d3f186648b4e8b76d6af0213a0cf3ec0c0fdafd61d7225/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-07-28 09:48:16.434000 [ 682 ] DEBUG : http://localhost:None "GET /v1.46/containers/037cfbebda22295370d3f186648b4e8b76d6af0213a0cf3ec0c0fdafd61d7225/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-07-28 09:48:16.537000 [ 682 ] DEBUG : http://localhost:None "GET /v1.46/containers/037cfbebda22295370d3f186648b4e8b76d6af0213a0cf3ec0c0fdafd61d7225/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-07-28 09:48:16.640000 [ 682 ] DEBUG : http://localhost:None "GET /v1.46/containers/037cfbebda22295370d3f186648b4e8b76d6af0213a0cf3ec0c0fdafd61d7225/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-07-28 09:48:16.743000 [ 682 ] DEBUG : http://localhost:None "GET /v1.46/containers/037cfbebda22295370d3f186648b4e8b76d6af0213a0cf3ec0c0fdafd61d7225/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-07-28 09:48:16.845000 [ 682 ] DEBUG : http://localhost:None "GET /v1.46/containers/037cfbebda22295370d3f186648b4e8b76d6af0213a0cf3ec0c0fdafd61d7225/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-07-28 09:48:16.948000 [ 682 ] DEBUG : http://localhost:None "GET /v1.46/containers/037cfbebda22295370d3f186648b4e8b76d6af0213a0cf3ec0c0fdafd61d7225/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-07-28 09:48:17.051000 [ 682 ] DEBUG : http://localhost:None "GET /v1.46/containers/037cfbebda22295370d3f186648b4e8b76d6af0213a0cf3ec0c0fdafd61d7225/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-07-28 09:48:17.153000 [ 682 ] DEBUG : http://localhost:None "GET /v1.46/containers/037cfbebda22295370d3f186648b4e8b76d6af0213a0cf3ec0c0fdafd61d7225/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-07-28 09:48:17.256000 [ 682 ] DEBUG : http://localhost:None "GET /v1.46/containers/037cfbebda22295370d3f186648b4e8b76d6af0213a0cf3ec0c0fdafd61d7225/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-07-28 09:48:17.359000 [ 682 ] DEBUG : http://localhost:None "GET /v1.46/containers/037cfbebda22295370d3f186648b4e8b76d6af0213a0cf3ec0c0fdafd61d7225/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-07-28 09:48:17.462000 [ 682 ] DEBUG : http://localhost:None "GET /v1.46/containers/037cfbebda22295370d3f186648b4e8b76d6af0213a0cf3ec0c0fdafd61d7225/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-07-28 09:48:17.565000 [ 682 ] DEBUG : http://localhost:None "GET /v1.46/containers/037cfbebda22295370d3f186648b4e8b76d6af0213a0cf3ec0c0fdafd61d7225/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-07-28 09:48:17.668000 [ 682 ] DEBUG : http://localhost:None "GET /v1.46/containers/037cfbebda22295370d3f186648b4e8b76d6af0213a0cf3ec0c0fdafd61d7225/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-07-28 09:48:17.771000 [ 682 ] DEBUG : http://localhost:None "GET /v1.46/containers/037cfbebda22295370d3f186648b4e8b76d6af0213a0cf3ec0c0fdafd61d7225/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-07-28 09:48:17.874000 [ 682 ] DEBUG : http://localhost:None "GET /v1.46/containers/037cfbebda22295370d3f186648b4e8b76d6af0213a0cf3ec0c0fdafd61d7225/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-07-28 09:48:17.976000 [ 682 ] DEBUG : http://localhost:None "GET /v1.46/containers/037cfbebda22295370d3f186648b4e8b76d6af0213a0cf3ec0c0fdafd61d7225/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-07-28 09:48:18.080000 [ 682 ] DEBUG : http://localhost:None "GET /v1.46/containers/037cfbebda22295370d3f186648b4e8b76d6af0213a0cf3ec0c0fdafd61d7225/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-07-28 09:48:18.183000 [ 682 ] DEBUG : http://localhost:None "GET /v1.46/containers/037cfbebda22295370d3f186648b4e8b76d6af0213a0cf3ec0c0fdafd61d7225/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-07-28 09:48:18.286000 [ 682 ] DEBUG : http://localhost:None "GET /v1.46/containers/037cfbebda22295370d3f186648b4e8b76d6af0213a0cf3ec0c0fdafd61d7225/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-07-28 09:48:18.389000 [ 682 ] DEBUG : http://localhost:None "GET /v1.46/containers/037cfbebda22295370d3f186648b4e8b76d6af0213a0cf3ec0c0fdafd61d7225/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-07-28 09:48:18.492000 [ 682 ] DEBUG : http://localhost:None "GET /v1.46/containers/037cfbebda22295370d3f186648b4e8b76d6af0213a0cf3ec0c0fdafd61d7225/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-07-28 09:48:18.595000 [ 682 ] DEBUG : http://localhost:None "GET /v1.46/containers/037cfbebda22295370d3f186648b4e8b76d6af0213a0cf3ec0c0fdafd61d7225/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-07-28 09:48:18.595000 [ 682 ] DEBUG : ClickHouse node started (cluster.py:3159, start) ------------------------------ Captured log call ------------------------------- 2025-07-28 09:48:18.598000 [ 682 ] DEBUG : Executing query CREATE USER IF NOT EXISTS A on node (cluster.py:3648, query) 2025-07-28 09:48:18.813000 [ 682 ] DEBUG : Executing query GRANT ALL ON *.* TO A on node (cluster.py:3648, query) 2025-07-28 09:48:19.028000 [ 682 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=1 on node (cluster.py:3716, get_query_request) 2025-07-28 09:48:19.029000 [ 682 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=80000000 on node (cluster.py:3716, get_query_request) 2025-07-28 09:48:19.030000 [ 682 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=1 on node (cluster.py:3716, get_query_request) 2025-07-28 09:48:19.031000 [ 682 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=80000000 on node (cluster.py:3716, get_query_request) 2025-07-28 09:48:19.031000 [ 682 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=1 on node (cluster.py:3716, get_query_request) 2025-07-28 09:48:19.032000 [ 682 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=80000000 on node (cluster.py:3716, get_query_request) 2025-07-28 09:48:19.034000 [ 682 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=1 on node (cluster.py:3716, get_query_request) 2025-07-28 09:48:19.035000 [ 682 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=80000000 on node (cluster.py:3716, get_query_request) 2025-07-28 09:48:19.036000 [ 682 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=1 on node (cluster.py:3716, get_query_request) 2025-07-28 09:48:19.037000 [ 682 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=80000000 on node (cluster.py:3716, get_query_request) 2025-07-28 09:48:19.038000 [ 682 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=1 on node (cluster.py:3716, get_query_request) 2025-07-28 09:48:19.038000 [ 682 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=80000000 on node (cluster.py:3716, get_query_request) 2025-07-28 09:48:19.039000 [ 682 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=1 on node (cluster.py:3716, get_query_request) 2025-07-28 09:48:19.040000 [ 682 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=80000000 on node (cluster.py:3716, get_query_request) 2025-07-28 09:48:19.041000 [ 682 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=1 on node (cluster.py:3716, get_query_request) 2025-07-28 09:48:19.044000 [ 682 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=80000000 on node (cluster.py:3716, get_query_request) 2025-07-28 09:48:19.055000 [ 682 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=1 on node (cluster.py:3716, get_query_request) 2025-07-28 09:48:19.078000 [ 682 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=80000000 on node (cluster.py:3716, get_query_request) 2025-07-28 09:48:19.092000 [ 682 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=1 on node (cluster.py:3716, get_query_request) 2025-07-28 09:48:19.101000 [ 682 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=80000000 on node (cluster.py:3716, get_query_request) 2025-07-28 09:48:19.116000 [ 682 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=1 on node (cluster.py:3716, get_query_request) 2025-07-28 09:48:19.128000 [ 682 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=80000000 on node (cluster.py:3716, get_query_request) 2025-07-28 09:48:19.140000 [ 682 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=1 on node (cluster.py:3716, get_query_request) 2025-07-28 09:48:19.196000 [ 682 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=80000000 on node (cluster.py:3716, get_query_request) 2025-07-28 09:48:19.216000 [ 682 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=1 on node (cluster.py:3716, get_query_request) 2025-07-28 09:48:19.230000 [ 682 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=80000000 on node (cluster.py:3716, get_query_request) 2025-07-28 09:48:19.236000 [ 682 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=1 on node (cluster.py:3716, get_query_request) 2025-07-28 09:48:19.241000 [ 682 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=80000000 on node (cluster.py:3716, get_query_request) 2025-07-28 09:48:19.241000 [ 682 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=1 on node (cluster.py:3716, get_query_request) 2025-07-28 09:48:19.251000 [ 682 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=80000000 on node (cluster.py:3716, get_query_request) 2025-07-28 09:48:19.270000 [ 682 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=1 on node (cluster.py:3716, get_query_request) 2025-07-28 09:48:19.286000 [ 682 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=80000000 on node (cluster.py:3716, get_query_request) 2025-07-28 09:48:19.296000 [ 682 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=1 on node (cluster.py:3716, get_query_request) 2025-07-28 09:48:19.305000 [ 682 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=80000000 on node (cluster.py:3716, get_query_request) 2025-07-28 09:48:19.318000 [ 682 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=1 on node (cluster.py:3716, get_query_request) 2025-07-28 09:48:19.344000 [ 682 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=80000000 on node (cluster.py:3716, get_query_request) 2025-07-28 09:48:19.355000 [ 682 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=1 on node (cluster.py:3716, get_query_request) 2025-07-28 09:48:19.366000 [ 682 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=80000000 on node (cluster.py:3716, get_query_request) 2025-07-28 09:48:19.377000 [ 682 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=1 on node (cluster.py:3716, get_query_request) 2025-07-28 09:48:19.380000 [ 682 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=80000000 on node (cluster.py:3716, get_query_request) 2025-07-28 09:48:19.399000 [ 682 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=1 on node (cluster.py:3716, get_query_request) 2025-07-28 09:48:19.413000 [ 682 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=80000000 on node (cluster.py:3716, get_query_request) 2025-07-28 09:48:19.431000 [ 682 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=1 on node (cluster.py:3716, get_query_request) 2025-07-28 09:48:19.440000 [ 682 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=80000000 on node (cluster.py:3716, get_query_request) 2025-07-28 09:48:19.442000 [ 682 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=1 on node (cluster.py:3716, get_query_request) 2025-07-28 09:48:19.444000 [ 682 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=80000000 on node (cluster.py:3716, get_query_request) 2025-07-28 09:48:19.471000 [ 682 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=1 on node (cluster.py:3716, get_query_request) 2025-07-28 09:48:19.490000 [ 682 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=80000000 on node (cluster.py:3716, get_query_request) 2025-07-28 09:48:19.491000 [ 682 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=1 on node (cluster.py:3716, get_query_request) 2025-07-28 09:48:19.501000 [ 682 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=80000000 on node (cluster.py:3716, get_query_request) 2025-07-28 09:48:19.508000 [ 682 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=1 on node (cluster.py:3716, get_query_request) 2025-07-28 09:48:19.516000 [ 682 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=80000000 on node (cluster.py:3716, get_query_request) 2025-07-28 09:48:19.523000 [ 682 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=1 on node (cluster.py:3716, get_query_request) 2025-07-28 09:48:19.533000 [ 682 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=80000000 on node (cluster.py:3716, get_query_request) 2025-07-28 09:48:19.547000 [ 682 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=1 on node (cluster.py:3716, get_query_request) 2025-07-28 09:48:19.609000 [ 682 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=80000000 on node (cluster.py:3716, get_query_request) 2025-07-28 09:48:19.610000 [ 682 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=1 on node (cluster.py:3716, get_query_request) 2025-07-28 09:48:19.620000 [ 682 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=80000000 on node (cluster.py:3716, get_query_request) 2025-07-28 09:48:19.631000 [ 682 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=1 on node (cluster.py:3716, get_query_request) 2025-07-28 09:48:19.640000 [ 682 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=80000000 on node (cluster.py:3716, get_query_request) 2025-07-28 09:48:19.643000 [ 682 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=1 on node (cluster.py:3716, get_query_request) 2025-07-28 09:48:19.644000 [ 682 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=80000000 on node (cluster.py:3716, get_query_request) 2025-07-28 09:48:19.647000 [ 682 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=1 on node (cluster.py:3716, get_query_request) 2025-07-28 09:48:19.651000 [ 682 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=80000000 on node (cluster.py:3716, get_query_request) 2025-07-28 09:48:19.653000 [ 682 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=1 on node (cluster.py:3716, get_query_request) 2025-07-28 09:48:19.659000 [ 682 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=80000000 on node (cluster.py:3716, get_query_request) 2025-07-28 09:48:19.667000 [ 682 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=1 on node (cluster.py:3716, get_query_request) 2025-07-28 09:48:19.688000 [ 682 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=80000000 on node (cluster.py:3716, get_query_request) 2025-07-28 09:48:19.698000 [ 682 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=1 on node (cluster.py:3716, get_query_request) 2025-07-28 09:48:19.701000 [ 682 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=80000000 on node (cluster.py:3716, get_query_request) 2025-07-28 09:48:19.718000 [ 682 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=1 on node (cluster.py:3716, get_query_request) 2025-07-28 09:48:19.729000 [ 682 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=80000000 on node (cluster.py:3716, get_query_request) 2025-07-28 09:48:19.740000 [ 682 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=1 on node (cluster.py:3716, get_query_request) 2025-07-28 09:48:19.762000 [ 682 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=80000000 on node (cluster.py:3716, get_query_request) 2025-07-28 09:48:19.778000 [ 682 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=1 on node (cluster.py:3716, get_query_request) 2025-07-28 09:48:19.802000 [ 682 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=80000000 on node (cluster.py:3716, get_query_request) 2025-07-28 09:48:19.804000 [ 682 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=1 on node (cluster.py:3716, get_query_request) 2025-07-28 09:48:19.814000 [ 682 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=80000000 on node (cluster.py:3716, get_query_request) 2025-07-28 09:48:19.824000 [ 682 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=1 on node (cluster.py:3716, get_query_request) 2025-07-28 09:48:19.836000 [ 682 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=80000000 on node (cluster.py:3716, get_query_request) 2025-07-28 09:48:19.837000 [ 682 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=1 on node (cluster.py:3716, get_query_request) 2025-07-28 09:48:19.848000 [ 682 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=80000000 on node (cluster.py:3716, get_query_request) 2025-07-28 09:48:19.864000 [ 682 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=1 on node (cluster.py:3716, get_query_request) 2025-07-28 09:48:19.867000 [ 682 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=80000000 on node (cluster.py:3716, get_query_request) 2025-07-28 09:48:19.875000 [ 682 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=1 on node (cluster.py:3716, get_query_request) 2025-07-28 09:48:19.888000 [ 682 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=80000000 on node (cluster.py:3716, get_query_request) 2025-07-28 09:48:19.895000 [ 682 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=1 on node (cluster.py:3716, get_query_request) 2025-07-28 09:48:19.896000 [ 682 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=80000000 on node (cluster.py:3716, get_query_request) 2025-07-28 09:48:19.901000 [ 682 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=1 on node (cluster.py:3716, get_query_request) 2025-07-28 09:48:19.903000 [ 682 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=80000000 on node (cluster.py:3716, get_query_request) 2025-07-28 09:48:19.916000 [ 682 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=1 on node (cluster.py:3716, get_query_request) 2025-07-28 09:48:19.926000 [ 682 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=80000000 on node (cluster.py:3716, get_query_request) 2025-07-28 09:48:19.935000 [ 682 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=1 on node (cluster.py:3716, get_query_request) 2025-07-28 09:48:19.951000 [ 682 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=80000000 on node (cluster.py:3716, get_query_request) 2025-07-28 09:48:19.962000 [ 682 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=1 on node (cluster.py:3716, get_query_request) 2025-07-28 09:48:19.977000 [ 682 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=80000000 on node (cluster.py:3716, get_query_request) 2025-07-28 09:48:19.982000 [ 682 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=1 on node (cluster.py:3716, get_query_request) 2025-07-28 09:48:19.992000 [ 682 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=80000000 on node (cluster.py:3716, get_query_request) 2025-07-28 09:48:20.005000 [ 682 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=1 on node (cluster.py:3716, get_query_request) 2025-07-28 09:48:20.013000 [ 682 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=80000000 on node (cluster.py:3716, get_query_request) ---------------------------- Captured log teardown ----------------------------- 2025-07-28 09:48:35.513000 [ 682 ] DEBUG : Command:[docker compose --env-file /ClickHouse/tests/integration/test_overcommit_tracker/_instances-1-gw4/.env --project-name roottestovercommittracker-gw4 --file /ClickHouse/tests/integration/test_overcommit_tracker/_instances-1-gw4/node/docker-compose.yml stop --timeout 20] (cluster.py:121, run_and_check) 2025-07-28 09:48:37.060000 [ 682 ] DEBUG : Stderr: Container roottestovercommittracker-gw4-node-1 Stopping (cluster.py:147, run_and_check) 2025-07-28 09:48:37.061000 [ 682 ] DEBUG : Stderr: Container roottestovercommittracker-gw4-node-1 Stopped (cluster.py:147, run_and_check) 2025-07-28 09:48:37.061000 [ 682 ] DEBUG : Command:[bash -c [ -f /ClickHouse/tests/integration/test_overcommit_tracker/_instances-1-gw4/node/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_overcommit_tracker/_instances-1-gw4/node/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] (cluster.py:121, run_and_check) 2025-07-28 09:48:37.074000 [ 682 ] DEBUG : Command:[docker compose --env-file /ClickHouse/tests/integration/test_overcommit_tracker/_instances-1-gw4/.env --project-name roottestovercommittracker-gw4 --file /ClickHouse/tests/integration/test_overcommit_tracker/_instances-1-gw4/node/docker-compose.yml down --volumes] (cluster.py:121, run_and_check) 2025-07-28 09:48:37.507000 [ 682 ] DEBUG : Stderr: Container roottestovercommittracker-gw4-node-1 Stopping (cluster.py:147, run_and_check) 2025-07-28 09:48:37.507000 [ 682 ] DEBUG : Stderr: Container roottestovercommittracker-gw4-node-1 Stopped (cluster.py:147, run_and_check) 2025-07-28 09:48:37.507000 [ 682 ] DEBUG : Stderr: Container roottestovercommittracker-gw4-node-1 Removing (cluster.py:147, run_and_check) 2025-07-28 09:48:37.507000 [ 682 ] DEBUG : Stderr: Container roottestovercommittracker-gw4-node-1 Removed (cluster.py:147, run_and_check) 2025-07-28 09:48:37.508000 [ 682 ] DEBUG : Stderr: Network roottestovercommittracker-gw4_default Removing (cluster.py:147, run_and_check) 2025-07-28 09:48:37.508000 [ 682 ] DEBUG : Stderr: Network roottestovercommittracker-gw4_default Removed (cluster.py:147, run_and_check) 2025-07-28 09:48:37.508000 [ 682 ] DEBUG : Cleanup called (cluster.py:851, cleanup) 2025-07-28 09:48:37.525000 [ 682 ] DEBUG : Docker networks for project roottestovercommittracker-gw4 are NETWORK ID NAME DRIVER SCOPE (cluster.py:830, print_all_docker_pieces) 2025-07-28 09:48:37.541000 [ 682 ] DEBUG : Docker containers for project roottestovercommittracker-gw4 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES (cluster.py:838, print_all_docker_pieces) 2025-07-28 09:48:37.557000 [ 682 ] DEBUG : Docker volumes for project roottestovercommittracker-gw4 are DRIVER VOLUME NAME (cluster.py:846, print_all_docker_pieces) 2025-07-28 09:48:37.558000 [ 682 ] DEBUG : Command:[docker container list --all --filter name='^/roottestovercommittracker-gw4-.*-1$' --format '{{.ID}}:{{.Names}}'] (cluster.py:121, run_and_check) 2025-07-28 09:48:37.576000 [ 682 ] DEBUG : Unstopped containers: {} (cluster.py:865, cleanup) 2025-07-28 09:48:37.576000 [ 682 ] DEBUG : No running containers for project: roottestovercommittracker-gw4 (cluster.py:879, cleanup) 2025-07-28 09:48:37.576000 [ 682 ] DEBUG : Trying to prune unused networks... (cluster.py:885, cleanup) 2025-07-28 09:48:37.594000 [ 682 ] DEBUG : Trying to prune unused images... (cluster.py:901, cleanup) 2025-07-28 09:48:37.594000 [ 682 ] DEBUG : Command:[docker image prune -f] (cluster.py:121, run_and_check) 2025-07-28 09:48:37.617000 [ 682 ] DEBUG : Stdout:Total reclaimed space: 0B (cluster.py:145, run_and_check) 2025-07-28 09:48:37.617000 [ 682 ] DEBUG : Images pruned (cluster.py:904, cleanup) 2025-07-28 09:48:37.617000 [ 682 ] DEBUG : Trying to prune unused volumes... (cluster.py:910, cleanup) 2025-07-28 09:48:37.617000 [ 682 ] DEBUG : Command:[docker volume ls | wc -l] (cluster.py:121, run_and_check) 2025-07-28 09:48:37.634000 [ 682 ] DEBUG : Stdout:3 (cluster.py:145, run_and_check) 2025-07-28 09:48:37.634000 [ 682 ] DEBUG : Command:[docker volume prune -f] (cluster.py:121, run_and_check) 2025-07-28 09:48:37.650000 [ 682 ] DEBUG : Stdout:Total reclaimed space: 0B (cluster.py:145, run_and_check) 2025-07-28 09:48:37.650000 [ 682 ] DEBUG : Volumes pruned: 3 (cluster.py:915, cleanup) ----------------- generated report log file: parallel1_1.jsonl ----------------- ============================== slowest durations =============================== 21.18s teardown test_merge_tree_check_part_with_cache/test.py::test_check_part_with_cache 20.87s setup test_fetch_memory_usage/test.py::test_huge_column 19.21s setup test_merge_tree_load_parts/test.py::test_merge_tree_load_parts_filesystem_error 17.51s setup test_hedged_requests_parallel/test.py::test_combination1 16.93s setup test_merge_tree_check_part_with_cache/test.py::test_check_part_with_cache 16.85s call test_overcommit_tracker/test.py::test_user_overcommit 14.40s setup test_overcommit_tracker/test.py::test_user_overcommit 4.10s teardown test_hedged_requests_parallel/test.py::test_send_table_status_sleep 3.61s teardown test_fetch_memory_usage/test.py::test_huge_column 2.66s teardown test_merge_tree_load_parts/test.py::test_merge_tree_load_parts_filesystem_error 2.14s teardown test_overcommit_tracker/test.py::test_user_overcommit 0.92s call test_jemalloc_percpu_arena/test.py::test_jemalloc_percpu_arena 0.37s call test_fetch_memory_usage/test.py::test_huge_column 0.32s call test_merge_tree_check_part_with_cache/test.py::test_check_part_with_cache 0.27s call test_hedged_requests_parallel/test.py::test_send_table_status_sleep 0.27s call test_merge_tree_load_parts/test.py::test_merge_tree_load_parts_filesystem_error 0.27s call test_hedged_requests_parallel/test.py::test_combination1 0.27s call test_hedged_requests_parallel/test.py::test_combination2 0.27s call test_hedged_requests_parallel/test.py::test_send_data 0.27s call test_hedged_requests_parallel/test.py::test_query_with_no_data_to_sample 0.04s setup test_jemalloc_percpu_arena/test.py::test_jemalloc_percpu_arena 0.00s teardown test_jemalloc_percpu_arena/test.py::test_jemalloc_percpu_arena 0.00s teardown test_hedged_requests_parallel/test.py::test_combination1 0.00s setup test_hedged_requests_parallel/test.py::test_send_table_status_sleep 0.00s teardown test_hedged_requests_parallel/test.py::test_send_data 0.00s setup test_hedged_requests_parallel/test.py::test_query_with_no_data_to_sample 0.00s setup test_hedged_requests_parallel/test.py::test_send_data 0.00s setup test_hedged_requests_parallel/test.py::test_combination2 0.00s teardown test_hedged_requests_parallel/test.py::test_query_with_no_data_to_sample 0.00s teardown test_hedged_requests_parallel/test.py::test_combination2 =========================== short test summary info ============================ FAILED test_overcommit_tracker/test.py::test_user_overcommit - AssertionError... SKIPPED [1] test_jemalloc_percpu_arena/test.py:62: Compiled without jemalloc (USE_JEMALLOC=b'') SKIPPED [1] test_merge_tree_check_part_with_cache/test.py:28: Skip with debug build and sanitizers. This test manually corrupts cache which triggers LOGICAL_ERROR and leads to crash with those builds SKIPPED [1] test_hedged_requests_parallel/test.py:199: Hedged requests don't work under Thread Sanitizer SKIPPED [1] test_hedged_requests_parallel/test.py:212: Hedged requests don't work under Thread Sanitizer SKIPPED [1] test_hedged_requests_parallel/test.py:226: Hedged requests don't work under Thread Sanitizer SKIPPED [1] test_hedged_requests_parallel/test.py:188: Hedged requests don't work under Thread Sanitizer SKIPPED [1] test_hedged_requests_parallel/test.py:176: Hedged requests don't work under Thread Sanitizer SKIPPED [1] test_merge_tree_load_parts/test.py:228: Skip with debug build and sanitizers. This test intentionally triggers LOGICAL_ERROR which leads to crash with those builds SKIPPED [1] test_fetch_memory_usage/test.py:26: sanitizer build has higher memory consumption; also it is slow ======================== 1 failed, 9 skipped in 40.37s ========================= Traceback (most recent call last): File "/home/ubuntu/_work/ClickHouse/ClickHouse/tests/integration/./runner", line 492, in subprocess.check_call(cmd, shell=True, bufsize=0) File "/usr/lib/python3.10/subprocess.py", line 369, in check_call raise CalledProcessError(retcode, cmd) subprocess.CalledProcessError: Command 'docker run --rm --name clickhouse_integration_tests_53fn1r --privileged --dns-search='.' --memory=30709018624 --security-opt seccomp=unconfined --cap-add=SYS_PTRACE --volume=/home/ubuntu/_work/_temp/test/build/clickhouse:/clickhouse --volume=/home/ubuntu/_work/ClickHouse/ClickHouse/programs/server:/clickhouse-config --volume=/home/ubuntu/_work/ClickHouse/ClickHouse/tests/integration:/ClickHouse/tests/integration --volume=/home/ubuntu/_work/ClickHouse/ClickHouse/utils/backupview:/ClickHouse/utils/backupview --volume=/home/ubuntu/_work/ClickHouse/ClickHouse/utils/grpc-client/pb2:/ClickHouse/utils/grpc-client/pb2 --volume=/run:/run/host:ro --volume=clickhouse_integration_tests_volume:/var/lib/docker -e DOCKER_DOTNET_CLIENT_TAG=11de0b29a15d -e DOCKER_HELPER_TAG=5dc43a6382f0 -e DOCKER_BASE_TAG=5ccda723c1fc -e DOCKER_KERBEROS_KDC_TAG=9391ecdee8d7 -e DOCKER_MYSQL_GOLANG_CLIENT_TAG=9bec2a638e6e -e DOCKER_MYSQL_JAVA_CLIENT_TAG=766bff31cfe4 -e DOCKER_MYSQL_JS_CLIENT_TAG=41ba7c2ec2a1 -e DOCKER_MYSQL_PHP_CLIENT_TAG=88be89c1e3b6 -e DOCKER_NGINX_DAV_TAG=b55ac9cd7519 -e DOCKER_POSTGRESQL_JAVA_CLIENT_TAG=a4eff5c7f4d6 -e DOCKER_PYTHON_BOTTLE_TAG=d862517635bf -e DOCKER_CLIENT_TIMEOUT=300 -e COMPOSE_HTTP_TIMEOUT=600 -e PYTHONUNBUFFERED=1 -e PYTEST_ADDOPTS="--dist=loadfile -n 10 -rfEps --run-id=1 --color=no --durations=0 --report-log=parallel1_1.jsonl --report-log-exclude-logs-on-passed-tests test_fetch_memory_usage/test.py::test_huge_column test_hedged_requests_parallel/test.py::test_combination1 test_hedged_requests_parallel/test.py::test_combination2 test_hedged_requests_parallel/test.py::test_query_with_no_data_to_sample test_hedged_requests_parallel/test.py::test_send_data test_hedged_requests_parallel/test.py::test_send_table_status_sleep test_jemalloc_percpu_arena/test.py::test_jemalloc_percpu_arena test_merge_tree_check_part_with_cache/test.py::test_check_part_with_cache test_merge_tree_load_parts/test.py::test_merge_tree_load_parts_filesystem_error test_overcommit_tracker/test.py::test_user_overcommit -vvv " altinityinfra/integration-tests-runner:226bfaf75ac1 ' returned non-zero exit status 1.